Expectation-Maximization via Pretext-Invariant Representations

نویسندگان

چکیده

Contrastive learning methods have been widely adopted in numerous unsupervised and self-supervised visual representation methods. Such algorithms aim to maximize the cosine similarity between two positive samples while minimizing that of negative samples. Recently, Grill et al. propose an algorithm, BYOL [1], utilize only samples, completely giving up on ones, by introducing a Siamese-like asymmetric architecture. Although many recent state-of-the-art (SOTA) adopt architecture, most them simply introduce additional neural network, predictor, without much exploration asymmetrical In contrast, He SimSiam [2], simple Siamese architecture relying stop-gradient operation instead momentum encoder describe framework from perspective Expectation-Maximization. We argue BYOL-like attain suboptimal performance due inconsistency during training. this work, we explain novel objective, Expectation-Maximization via Pretext-Invariant Representations (EMPIR), which enhances Expectation-Maximization-based optimization enforcing augmentation invariance within local region k nearest neighbors, resulting consistent learning. other words, as core task architectures. show it consistently outperforms decent margin. also demonstrate its transfer capabilities downstream image recognition tasks.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Expectation Maximization

The Expectation Maximization (EM) algorithm [1, 2] is one of the most widely used algorithms in statistics. Suppose we are given some observed data X and a model family parametrized by θ, and would like to find the θ which maximizes p(X |θ), i.e. the maximum likelihood estimator. The basic idea of EM is actually quite simple: when direct maximization of p(X |θ) is complicated we can augment the...

متن کامل

Week 9: Expectation Maximization

Last week, we saw how we could represent clustering with a probabilistic model. In this model, called a Gaussian mixture model, we model each datapoint x i as originating from some cluster, with a corresponding cluster label y i distributed according to p(y), and the corresponding distribution for that cluster given by a multivariate Gaussian: p(x|y = k) =

متن کامل

Relational Neural Expectation Maximization

Common-sense physical reasoning is an essential ingredient for any intelligent agent operating in the real-world. For example, it can be used to simulate the environment, or to infer the state of parts of the world that are currently unobserved. In order to match real-world conditions this causal knowledge must be learned without access to supervised data. To solve this problem, we present a no...

متن کامل

Neural Expectation Maximization

We introduce a novel framework for clustering that combines generalized EM with neural networks and can be implemented as an end-to-end differentiable recurrent neural network. It learns its statistical model directly from the data and can represent complex non-linear dependencies between inputs. We apply our framework to a perceptual grouping task and empirically verify that it yields the inte...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE Access

سال: 2023

ISSN: ['2169-3536']

DOI: https://doi.org/10.1109/access.2023.3289589